# Cross-lingual transfer

Gender Prediction Model From Text
MIT
This model is built on DeBERTa-v3-large and can predict the gender of anonymous speakers or authors based on the content of English texts.
Text Classification Transformers English
G
fc63
106
1
Moritzlaurer Mdeberta V3 Base Mnli Xnli
MIT
Natural language inference model supporting 100 languages, suitable for multilingual zero-shot classification tasks
Text Classification Supports Multiple Languages
M
MaagDeveloper
47
0
USER2 Small
Apache-2.0
USER2 is a next-generation Russian universal sentence encoder, specifically designed to support long-context sentence representations of up to 8,192 tokens.
Text Embedding Other
U
deepvk
1,409
6
USER Bge M3
Apache-2.0
Russian universal sentence encoder, based on the sentence-transformers framework, specifically designed to extract 1024-dimensional dense vectors for Russian text
Text Embedding Other
U
deepvk
339.46k
58
Llama 3 Typhoon V1.5x 8b Instruct
An 8-billion-parameter instruction model specifically designed for Thai, with performance comparable to GPT-3.5-turbo, optimized for application scenarios, retrieval-augmented generation, constrained generation, and reasoning tasks
Large Language Model Transformers Supports Multiple Languages
L
scb10x
3,269
16
Labse Ru Sts
MIT
High-quality Russian sentence embedding BERT model, optimized based on cointegrated/LaBSE-en-ru, suitable for semantic text similarity tasks
Text Embedding Transformers Other
L
sergeyzh
4,650
6
Multilingual Xlm Roberta For Ner
MIT
A named entity recognition model fine-tuned based on the XLM-RoBERTa base model, supporting multiple languages and capable of identifying three types of entities: locations, organizations, and persons.
Sequence Labeling Transformers
M
Tirendaz
56
2
Eng Guj Sim2b
An English - Gujarati translation model fine-tuned based on facebook/mbart-large-50-many-to-many-mmt
Machine Translation Transformers
E
hopkins
4
0
Biencoder Electra Base Mmarcofr
MIT
This is a dense single-vector dual-encoder model for French, designed for semantic search. The model maps queries and passages to 768-dimensional dense vectors and calculates relevance through cosine similarity.
Text Embedding French
B
antoinelouis
31
2
Ernie M Large Mnli Xnli
Apache-2.0
Natural language inference model supporting 100 languages, suitable for multilingual zero-shot classification tasks
Large Language Model Transformers Supports Multiple Languages
E
MoritzLaurer
94
18
Multilingual MiniLMv2 L12 Mnli Xnli
MIT
A multilingual natural language inference model supporting over 100 languages, suitable for zero-shot classification tasks
Text Classification Transformers Supports Multiple Languages
M
MoritzLaurer
245
6
Xlm V Base Mnli Xnli
MIT
A multilingual natural language inference model based on XLM-V-base, supporting zero-shot classification tasks in 116 languages
Large Language Model Transformers Supports Multiple Languages
X
MoritzLaurer
293
22
Mdeberta V3 Base Xnli Multilingual Nli 2mil7
MIT
Multilingual natural language inference model based on mDeBERTa-v3-base, supporting zero-shot classification in 100 languages
Large Language Model Transformers Supports Multiple Languages
M
MoritzLaurer
186.62k
316
Mbart Large 50 Finetuned V1
A fine-tuned model based on the mbart-large-50 architecture, suitable for multilingual summarization tasks
Large Language Model Transformers
M
z-rahimi-r
14
0
Xlm Roberta Base Finetuned Panx De Fr
MIT
A cross-lingual model fine-tuned on German and French datasets based on XLM-RoBERTa-base, primarily used for named entity recognition tasks.
Large Language Model Transformers
X
andreaschandra
15
0
Xlm Roberta Base Finetuned Panx All
MIT
Named entity recognition model fine-tuned on multilingual datasets based on xlm-roberta-base
Large Language Model Transformers
X
flood
15
0
Xlm Roberta Base Finetuned Panx De Fr
MIT
Cross-lingual model fine-tuned on German and French datasets based on XLM-RoBERTa-base
Large Language Model Transformers
X
skr3178
15
0
Hiner Original Xlm Roberta Large
This model is a named entity recognition (NER) model trained on the HiNER-original dataset based on the XLM-RoBERTa-large architecture, specifically designed for token classification tasks.
Sequence Labeling Transformers
H
cfilt
56
1
Roberta Large Wechsel Ukrainian
MIT
RoBERTa-large model migrated to Ukrainian using the WECHSEL method, excelling in NER and POS tagging tasks
Large Language Model Transformers Other
R
benjamin
75
0
Roberta Finetuned Ner
MIT
Named Entity Recognition (NER) model fine-tuned based on xlm-roberta-base, demonstrating excellent performance on the evaluation set (F1 score 0.9777)
Sequence Labeling Transformers
R
kSaluja
25
0
Multilingual Bert Base Cased Spanish
This model enhances the performance of low-resource language QA systems using English data through cascading adapter technology.
Question Answering System Transformers
M
bhavikardeshna
15
0
Tner Xlm Roberta Base Uncased Ontonotes5
This is an XLM-RoBERTa model fine-tuned for named entity recognition tasks, suitable for entity recognition in multilingual texts.
Sequence Labeling Transformers
T
asahi417
605
1
Roberta Base Wechsel German
MIT
A German RoBERTa model trained using the WECHSEL method, achieving cross-lingual transfer of monolingual language models through effective initialization of subword embeddings.
Large Language Model Transformers German
R
benjamin
96
7
Xlm Roberta Base Ft Udpos28 Ro
Apache-2.0
A multilingual POS tagging model based on XLM-RoBERTa, fine-tuned on Universal Dependencies v2.8 dataset with special optimization for Romanian
Sequence Labeling Transformers Other
X
wietsedv
14
1
Unispeech 1350 En 353 Fr Ft 1h
UniSpeech is a unified speech representation learning model that combines labeled and unlabeled data for pre-training, specifically fine-tuned for French.
Speech Recognition Transformers French
U
microsoft
20
0
Gpt2 Wechsel French
MIT
A French version of GPT-2 trained using the WECHSEL method, achieving cross-lingual transfer of monolingual language models through effective initialization of subword embeddings.
Large Language Model Transformers French
G
benjamin
33
4
Byt5 Xl
Apache-2.0
ByT5 is Google's token-free version of T5 that directly processes raw UTF-8 bytes, supporting multilingual text processing with robustness to noisy text.
Large Language Model Transformers Supports Multiple Languages
B
google
334
12
Roberta Base Wechsel Swahili
MIT
A RoBERTa base model trained using the WECHSEL method, specifically optimized for Swahili to achieve efficient cross-lingual transfer.
Large Language Model Transformers Other
R
benjamin
222
1
Xlm Roberta Base Ft Udpos28 En
Apache-2.0
This model is a multilingual POS tagging model based on the XLM-RoBERTa architecture, fine-tuned on the Universal Dependencies v2.8 dataset.
Sequence Labeling Transformers English
X
wietsedv
20
2
Wav2vec2 Large Xlsr 53 Kalmyk
Apache-2.0
This is a Kalmyk automatic speech recognition model based on the Wav2Vec2 architecture, pre-trained and fine-tuned to support Kalmyk speech-to-text tasks.
Speech Recognition Transformers Other
W
tugstugi
79
0
Bert Multi Cased Finetuned Xquadv1
Based on Google's BERT base multilingual model, fine-tuned on Q&A datasets in 11 languages, supporting cross-lingual Q&A tasks
Question Answering System Other
B
mrm8488
1,100
5
Xlm Roberta Base Ft Udpos28 Sk
Apache-2.0
Multilingual POS tagging model based on XLM-RoBERTa, specifically optimized for Slovak
Sequence Labeling Transformers Other
X
wietsedv
14
1
Unispeech 1350 En 168 Es Ft 1h
UniSpeech is a unified speech representation learning model that combines labeled and unlabeled data for pre-training, specifically fine-tuned for Spanish phoneme recognition.
Speech Recognition Transformers Spanish
U
microsoft
19
0
Multilingual Bert Base Cased Vietnamese
This model improves the performance of low-resource language question-answering systems by leveraging English data through cascading adapter technology.
Question Answering System Transformers
M
bhavikardeshna
31
0
Roberta Base Wechsel Chinese
MIT
A RoBERTa Chinese model trained with the WECHSEL method, achieving efficient cross-lingual transfer from English to Chinese
Large Language Model Transformers Chinese
R
benjamin
16
2
Xlm Roberta Base Ft Udpos28 Got
Apache-2.0
A multilingual POS tagging model based on XLM-RoBERTa, specifically optimized for Gothic
Sequence Labeling Transformers Other
X
wietsedv
15
0
Multilingual MiniLM L12 H384
MIT
MiniLM is a compact and efficient pre-trained language model that compresses Transformer models through deep self-attention distillation technology, supporting multilingual understanding and generation tasks.
Large Language Model Supports Multiple Languages
M
microsoft
28.51k
83
Xlm Roberta Base Arabic
This model improves the performance of low-resource language QA systems using English data through cascading adapter technology.
Question Answering System Transformers
X
bhavikardeshna
61
1
Gpt2 Wechsel Chinese
MIT
A Chinese GPT-2 model trained using the WECHSEL method, achieving cross-lingual transfer of monolingual language models through effective initialization of subword embeddings.
Large Language Model Transformers Chinese
G
benjamin
19
4
Mbart Large 50 Many To Many Mmt
A multilingual machine translation model fine-tuned based on mBART-large-50, supporting translation between 50 languages
Machine Translation Supports Multiple Languages
M
facebook
404.66k
357
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase